# Text generation optimization
Deepseek R1 Distill Qwen 14B GRPO Taiwan Spirit
This is a fine-tuned version based on the Qwen-14B model, trained using the GRPO method, suitable for text generation tasks.
Large Language Model
Transformers

D
kartd
111
1
Arshstory
MIT
A 500-million-parameter text generation model based on the Llama architecture, specifically designed for story creation.
Text Generation
Transformers

A
arshiaafshani
131
1
Latitudegames.muse 12B GGUF
Muse-12B is a text generation model with 12B parameters, developed by LatitudeGames, aiming to provide high-quality text generation capabilities.
Large Language Model
L
DevQuasar
291
1
Magtie V1 12B
Apache-2.0
MagTie-v1-12B is a 12B-parameter language model merged using the DARE TIES algorithm, combining the strengths of multiple pre-trained models
Large Language Model
Transformers

M
grimjim
32
2
Mistral Small 24B Instruct 2501 GGUF
Apache-2.0
Mistral-Small-24B-Instruct-2501 is a 24B-parameter instruction-finetuned large language model supporting multilingual text generation tasks.
Large Language Model Supports Multiple Languages
M
bartowski
48.61k
111
Salesforce.llama Xlam 2 70b Fc R GGUF
Llama-xLAM-2-70b-fc-r is a large language model released by Salesforce, based on the Llama 2 architecture with 70 billion parameters.
Large Language Model
S
DevQuasar
238
1
MT Gen10 Gemma 2 9B
This is a multi-model fusion version based on the Gemma-2-9B series models, merged using the DARE TIES method, integrating the strengths of multiple Gemma variants.
Large Language Model
Transformers

M
zelk12
26
2
Mtmme Merge Gemma 2 9B
A text generation model merged from Gemma-2B and Gemma-9B models using the SLERP method
Large Language Model
Transformers

M
zelk12
19
2
Irix 12B Model Stock
This is a result of merging multiple 12B-parameter-scale language models using the mergekit tool through the model inventory method
Large Language Model
Transformers

I
DreadPoor
373
9
L3.3 Cu Mai R1 70b
A 70B-parameter large language model based on the Llama3 architecture, specially optimized
Large Language Model
L
Steelskull
164
14
Llama 3.2 1B Instruct GGUF
The GGUF format version of Llama-3.2-1B-Instruct, providing broader support and better performance.
Large Language Model
L
MaziyarPanahi
190.76k
12
Mistral NeMo Minitron 8B Base IMat GGUF
Other
This is the result of llama.cpp imatrix quantization based on the nvidia/Mistral-NeMo-Minitron-8B-Base model, providing more options for model usage and deployment.
Large Language Model
M
legraphista
1,115
1
Wizardlm 2 7B Abliterated GGUF
Apache-2.0
Llamacpp imatrix quantized version of WizardLM-2-7B-abliterated, offering multiple quantization options for different hardware configurations.
Large Language Model
W
bartowski
2,561
13
Lola V1
LOLA is an ultra-large-scale multilingual large model based on the sparse Mixture-of-Experts (MoE) Transformer architecture, supporting over 160 languages, with competitive advantages in natural language generation and understanding tasks.
Large Language Model
Transformers Other

L
dice-research
867
10
Fusellm 7B
Apache-2.0
FuseLLM-7B is a unified model that integrates knowledge from multiple open-source large language models, combining the capabilities of LLMs with different architectures through knowledge fusion technology.
Large Language Model
Transformers Supports Multiple Languages

F
Wanfq
45
22
Caplattessdolxaboros Yi 34B 200K DARE Ties HighDensity
Other
This is a high-density merged model based on the Yi-34B-200K foundation model, integrating multiple homologous models through the DARE Ties method, featuring 200K long-context processing capability.
Large Language Model
Transformers English

C
brucethemoose
94
12
Tiny Vicuna 1B GGUF
Tiny-Vicuna-1B is a lightweight model fine-tuned from TinyLLama 1.1B using the WizardVicuna dataset, designed for early-stage experimental iterations.
Large Language Model
T
afrideva
208.74k
6
Spelling Correction Multilingual Base
MIT
An experimental model for correcting spelling errors and punctuation in English and German
Text Generation
Transformers Supports Multiple Languages

S
oliverguhr
655
11
Flan T5 Xxl Sharded Fp16
Apache-2.0
FLAN-T5 XXL is a variant of Google's T5 model, fine-tuned on over 1,000 additional tasks, supports multiple languages, and outperforms the original T5 model.
Large Language Model
Transformers

F
philschmid
531
54
Reward Model Deberta V3 Large
MIT
This reward model is trained to predict which generated answer human evaluators would prefer for a given question.
Large Language Model
Transformers English

R
OpenAssistant
796
23
Gpt2 Medium Dutch Embeddings
A Dutch language model based on the medium-scale GPT-2 version, with only the vocabulary embedding layer retrained for Dutch adaptation.
Large Language Model Other
G
GroNLP
27
3
Featured Recommended AI Models